Bayesian Learning for Deep Neural Network Adaptation

نویسندگان

چکیده

A key task for speech recognition systems is to reduce the mismatch between training and evaluation data that often attributable speaker differences. Speaker adaptation techniques play a vital role mismatch. Model-based approaches require sufficient amounts of target ensure robustness. When amount level limited, prone overfitting poor generalization. To address issue, this paper proposes full Bayesian learning based DNN framework model speaker-dependent (SD) parameter uncertainty given limited specific data. This investigated in three forms techniques: hidden unit contributions (BLHUC), parameterized activation functions (BPAct), bias vectors (BHUB). In methods, deterministic SD parameters are replaced by latent variable posterior distributions each speaker, whose efficiently estimated using variational inference approach. Experiments conducted on 300-hour speed perturbed Switchboard corpus trained LF-MMI TDNN/CNN-TDNN suggest proposed consistently outperform NIST Hub5'00 RT03 sets. only first five utterances from as data, significant word error rate reductions up 1.4% absolute (7.2% relative) were obtained CallHome subset. The efficacy further demonstrated comparison against state-of-the-art performance same most recent reported literature.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Bayesian Incremental Learning for Deep Neural Networks

In industrial machine learning pipelines, data often arrive in parts. Particularly in the case of deep neural networks, it may be too expensive to train the model from scratch each time, so one would rather use a previously learned model and the new data to improve performance. However, deep neural networks are prone to getting stuck in a suboptimal solution when trained on only new data as com...

متن کامل

Using Deep Neural Network Approximate Bayesian Network

We present a new method to approximate posterior probabilities of Bayesian Network using Deep Neural Network. Experiment results on several public Bayesian Network datasets shows that Deep Neural Network is capable of learning joint probability distribution of Bayesian Network by learning from a few observation and posterior probability distribution pairs with high accuracy. Compared with tradi...

متن کامل

Reinforced backpropagation for deep neural network learning

Standard error backpropagation is used in almost all modern deep network training. However, it typically suffers from proliferation of saddle points in high-dimensional parameter space. Therefore, it is highly desirable to design an efficient algorithm to escape from these saddle points and reach a good parameter region of better generalization capabilities, especially based on rough insights a...

متن کامل

Learning Neural Network with Learning Rate Adaptation

In this chapter the analog VLSI implementation of a Multi Layer Perceptron (MLP) network with on-chip learning capability is presented. A MLP architecture is chosen since it can be applied to successfully solve real-world tasks, see among others [33, 6, 9,4]. Many examples of analog implementations of neural networks with on-chip learning capability have been presented in literature, for exampl...

متن کامل

Rapid adaptation for deep neural networks through multi-task learning

We propose a novel approach to addressing the adaptation effectiveness issue in parameter adaptation for deep neural network (DNN) based acoustic models for automatic speech recognition by adding one or more small auxiliary output layers modeling broad acoustic units, such as mono-phones or tied-state (often called senone) clusters. In scenarios with a limited amount of available adaptation dat...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE/ACM transactions on audio, speech, and language processing

سال: 2021

ISSN: ['2329-9304', '2329-9290']

DOI: https://doi.org/10.1109/taslp.2021.3084072